Error bounds for monomial convexification in polynomial optimization
نویسندگان
چکیده
Convex hulls of monomials have been widely studied in the literature, and monomial convexifications are implemented in global optimization software for relaxing polynomials. However, there has been no study of the error in the global optimum from such approaches. We give bounds on the worst-case error for convexifying a monomial over subsets of [0, 1]. This implies additive error bounds for relaxing a polynomial optimization problem by convexifying each monomial separately. Our main error bounds depend primarily on the degree of the monomial, making them easy to compute. Since monomial convexification studies depend on the bounds on the associated variables, in the second part, we conduct an error analysis for a multilinear monomial over two different types of box constraints. As part of this analysis, we also derive the convex hull of a multilinear monomial over [−1, 1].
منابع مشابه
Lp Relaxations Better than Convexification for Multicommodity Network Optimization Problems with Step Increasing Cost Functions
We address here a class of particularly hard-to-solve combinatorial optimization problems namely multicommodity network optimization when the link cost functions are discontinuous step increasing. The main focus is on the development of relaxations for such problems in order to derive lower bounds. A straightforward way of getting lower bounds is to solve the “convexification” of the problem, i...
متن کاملMonomial-wise optimal separable underestimators for mixed-integer polynomial optimization
In this paper we introduce a new method for solving box-constrained mixed-integer polynomial problems to global optimality. The approach, a specialized branch-and-bound algorithm, is based on the computation of lower bounds provided by the minimization of separable underestimators of the polynomial objective function. The underestimators are the novelty of the approach because the standard appr...
متن کاملNew Bases for Polynomial-Based Spaces
Since it is well-known that the Vandermonde matrix is ill-conditioned, while the interpolation itself is not unstable in function space, this paper surveys the choices of other new bases. These bases are data-dependent and are categorized into discretely l2-orthonormal and continuously L2-orthonormal bases. The first one construct a unitary Gramian matrix in the space l2(X) while the late...
متن کاملBackward Error of Polynomial Eigenvalue Problems Solved by Linearization of Lagrange Interpolants
This article considers the backward error of the solution of polynomial eigenvalue problems expressed as Lagrange interpolants. One of the most common strategies to solve polynomial eigenvalue problems is to linearize, which is to say that the polynomial eigenvalue problem is transformed into an equivalent larger linear eigenvalue problem, and solved using any appropriate eigensolver. Much of t...
متن کاملError Bounds for Some Semidefinite Programming Approaches to Polynomial Minimization on the Hypercube
We consider the problem of minimizing a polynomial on the hypercube [0, 1] and derive new error bounds for the hierarchy of semidefinite programming approximations to this problem corresponding to the Positivstellensatz of Schmüdgen [26]. The main tool we employ is Bernstein approximations of polynomials, which also gives constructive proofs and degree bounds for positivity certificates on the ...
متن کامل